Fusing In-Storage and Near-Storage Acceleration of Convolutional Neural Networks
نویسندگان
چکیده
Video analytics have a wide range of applications and has attracted much interest over the years. While it can be both computationally energy intensive, video greatly benefit from in/ near memory compute. The practice moving compute closer to continued show improvements performance consumption is seeing increasing adoption. Recent advancements in solid state drives (SSDs) incorporated Field Programmable Gate Arrays (FPGAs) with shared access drive’s storage cells. These FPGAs are capable running operations required by analytic pipelines such as object detection template matching. typically executed using Convolutional Neural Networks (CNNs). A CNN composed multiple individually processed layers which perform various image processing tasks. Due lack resources, layer may partitioned into more manageable sub-layers. sub-layers then sequentially, however some simultaneously. Moreover, cells within FPGA equipped SSD’s being augmented in-storage accelerate workloads exploit intra parallelism layer. To this end we present our work, leverages heterogeneous architectures create an in/near-storage acceleration solution for analytics. We designed NAND flash accelerator, mapped evaluated several benchmarks. how utilize FPGAs, local DRAMs, in-memory SSD workloads. Our work also demonstrates remove unnecessary transfers save latency energy.
منابع مشابه
Hardware Acceleration of Convolutional Neural Networks
Convolutional neural networks have been widely employed for image recognition applications because of their high accuracy, which they achieve by emulating how our own brain recognizes objects. The possibility of making our electronic devices recognize their surroundings have spawned a vast number potential of useful applications, including video surveillance, mobile robot vision, image search i...
متن کاملFlattened Convolutional Neural Networks for Feedforward Acceleration
We present flattened convolutional neural networks that are designed for fast feedforward execution. The redundancy of the parameters, especially weights of the convolutional filters in convolutional neural networks has been extensively studied and different heuristics have been proposed to construct a low rank basis of the filters after training. In this work, we train flattened networks that ...
متن کاملConstruction cost estimation of spherical storage tanks: artificial neural networks and hybrid regression—GA algorithms
One of the most important processes in the early stages of construction projects is to estimate the cost involved. This process involves a wide range of uncertainties, which make it a challenging task. Because of unknown issues, using the experience of the experts or looking for similar cases are the conventional methods to deal with cost estimation. The current study presents data-driven metho...
متن کاملConvolutional Analysis Operator Learning: Acceleration, Convergence, Application, and Neural Networks
Convolutional operator learning is increasingly gaining attention in many signal processing and computer vision applications. Learning kernels has mostly relied on so-called local approaches that extract and store many overlapping patches across training signals. Due to memory demands, local approaches have limitations when learning kernels from large datasets—particularly with multi-layered st...
متن کاملHENet:A Highly Efficient Convolutional Neural Networks Optimized for Accuracy, Speed and Storage
In order to enhance the real-time performance of convolutional neural networks (CNN), more and more researchers are focusing on improving the efficiency of CNN. Based on the analysis of some CNN architectures, such as ResNet, DenseNet, ShuffleNet and so on, we combined their advantages and proposed a very efficient model called Highly Efficient Networks(HENet). The new architecture uses an unus...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: ACM Journal on Emerging Technologies in Computing Systems
سال: 2023
ISSN: ['1550-4832', '1550-4840']
DOI: https://doi.org/10.1145/3597496